N-best rescoring by phoneme classifiers using subclass adaboost algorithm
نویسندگان
چکیده
This paper proposes a novel technique to exploit discriminative models with subclasses for speech recognition. Speech recognition using discriminative models has attracted much attention in the past decade. However, most discriminative models are still based on tree clustering results of HMM states. On the contrary, our proposed method, referred to as subclass AdaBoost, jointly selects optimal data split and weak discriminators in each iteration of the training process, and forms a weak classifier as a composite of these weak discriminators. As a result, a strong discriminator robust to a variety of subclasses is constructed without explicit clustering in advance. In the experiment, the subclass AdaBoost is applied to phoneme classification, and N-best hypotheses are rescored using the subclass AdaBoost phoneme classifiers. Experimental results show that the proposed method reduces word errors by over 10% relatively in a continuous speech recognition task.
منابع مشابه
Extending AdaBoost to Iteratively Vary Its Base Classifiers
This paper introduces AdaBoost Dynamic, an extension of AdaBoost.M1 algorithm by Freund and Shapire. In this extension we use different “weak” classifiers in subsequent iterations of the algorithm, instead of AdaBoost’s fixed base classifier. The algorithm is tested with various datasets from UCI database, and results show that the algorithm performs equally well as AdaBoost with the best possi...
متن کاملImproving WFST-based G2P Conversion with Alignment Constraints and RNNLM N-best Rescoring
This work introduces a modified WFST-based multiple to multiple EM-driven alignment algorithm for Grapheme-to-Phoneme (G2P) conversion, and preliminary experimental results applying a Recurrent Neural Network Language Model (RNNLM) as an Nbest rescoring mechanism for G2P conversion. The alignment algorithm leverages the WFST framework and introduces several simple structural constraints which y...
متن کاملUsing boosting to improve a hybrid HMM/neural network speech recognizer
”Boosting” is a general method for improving the performance of almost any learning algorithm. A recently proposed and very promising boosting algorithm is AdaBoost [7]. In this paper we investigate if AdaBoost can be used to improve a hybrid HMM/ neural network continuous speech recognizer. Boosting significantly improves the word error rate from 6.3% to 5.3% on a test set of the OGI Numbers95...
متن کاملAccelerating AdaBoost using UCB
This paper explores how multi-armed bandits (MABs) can be applied to accelerate AdaBoost. AdaBoost constructs a strong classifier in a stepwise fashion by adding simple base classifiers to a pool and using their weighted “vote” to determine the final classification. We model this stepwise base classifier selection as a sequential decision problem, and optimize it with MABs. Each arm represents ...
متن کاملOne-Pass Boosting
This paper studies boosting algorithms that make a single pass over a set of base classifiers. We first analyze a one-pass algorithm in the setting of boosting with diverse base classifiers. Our guarantee is the same as the best proved for any boosting algorithm, but our one-pass algorithm is much faster than previous approaches. We next exhibit a random source of examples for which a “picky” v...
متن کامل